A Bayesian view of language evolution by iterated learning

نویسندگان

  • Thomas L. Griffiths
  • Michael L. Kalish
چکیده

Models of language evolution have demonstrated how aspects of human language, such as compositionality, can arise in populations of interacting agents. This paper analyzes how languages change as the result of a particular form of interaction: agents learning from one another. We show that, when the learners are rational Bayesian agents, this process of iterated learning converges to the prior distribution over languages assumed by those learners. The rate of convergence is set by the amount of information conveyed by the data seen by each generation; the less informative the data, the faster the process converges to the prior. Human languages form a subset of all logically possible communication schemes, with universal properties shared by all languages (Comrie, 1981; Greenberg, 1963; Hawkins, 1988). A traditional explanation for these linguistic universals is that they are the consequence of constraints on the set of learnable languages imposed by an innate, language-specific, genetic endowment (e.g., Chomsky, 1965). Recent research has explored an alternative explanation: that universals emerge from evolutionary processes produced by the transmission of languages across generations (e.g., Kirby, 2001; Nowak, Plotkin, & Jansen, 2000). Languages change as each generation learns from that which preceded it. This process of iterated learning implicitly selects for languages that are more learnable. This suggests a tantalizing hypothesis: that iterated learning might be sufficient to explain the emergence of linguistic universals (Briscoe, 2002). Kirby (2001) introduced a framework for exploring this hypothesis, called the iterated learning model (ILM). In the ILM, each generation consists of one or more learners. Each learner sees some data, forms a hypothesis about the process that produced that data, and then produces the data which will be supplied to the next generation of learners, as shown in Figure 1 (a). The languages that succeed in being transmitted across generations are those that pass through the “information bottleneck” imposed by iterated learning. If particular properties of languages make it easier to pass through that bottleneck, then many generations of iterated learning might allow those properties to become universal. The ILM can be used to explore how different assumptions about language learning influence language evolution. A variety of learning algorithms have been examined using the ILM, including a heuristic grammar inducer (Kirby, 2001), associative networks (Smith, Kirby, & Brighton, 2003), and minimum description length (Brighton, 2002). Iterated learning with these algorithms produces languages that possess one of the most compelling properties of human languages: compositionality. In a compositional language, the meaning of an utterance is a function of the meaning of its parts. The intuitive explanation for these results is that the regular structure of compositional languages means that they can be learned from less data, and are thus more likely to pass through the information bottleneck. These instances of compositionality emerging from iterated learning raise an important question: what languages will survive many generations of iterated learning? While the circumstances under which compositionality will emerge from iterated learning with specific learning algorithms have been investigated (Brighton, 2002; Smith et al., 2003), there are no general results for arbitrary properties of languages or broad classes of learning algorithms. In this paper, we analyze iterated learning for the case where the learners are rational Bayesian agents. A variety of learning algorithms can be formulated in terms of Bayesian inference, and Bayesian methods underlie many approaches in computational linguistics (Manning & Schütze, 1999). The assumption that the learners are Bayesian agents makes it possible to derive analytic results indicating which languages will be favored by iterated learning. In particular, we prove the surprising result that the probability distribution over languages resulting from iterated Bayesian learning converges to the prior probability distribution assumed by the learners. This implies that the asymptotic probability that a language is used does not depend at all upon the properties of the language, being determined entirely by the assumptions of the learner. hypothesis data hypothesis data (a) g en er a ti o n

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Thomas’ theorem meets Bayes’ rule: a model of the iterated learning of language

We develop a Bayesian Iterated Learning Model (BILM) that models the cultural evolution of language as it is transmitted over generations of learners. We study the outcome of iterated learning in relation to the behavior of individual agents (their biases) and the social structure through which they transmit their behavior. BILM makes individual learning biases explicit and offers a direct comp...

متن کامل

Language Evolution by Iterated Learning With Bayesian Agents

Languages are transmitted from person to person and generation to generation via a process of iterated learning: people learn a language from other people who once learned that language themselves. We analyze the consequences of iterated learning for learning algorithms based on the principles of Bayesian inference, assuming that learners compute a posterior distribution over languages by combi...

متن کامل

The evolution of frequency distributions: relating regularization to inductive biases through iterated learning.

The regularization of linguistic structures by learners has played a key role in arguments for strong innate constraints on language acquisition, and has important implications for language evolution. However, relating the inductive biases of learners to regularization behavior in laboratory tasks can be challenging without a formal model. In this paper we explore how regular linguistic structu...

متن کامل

Convergence Bounds for Language Evolution by Iterated Learning

Similarities between human languages are often taken as evidence of constraints on language learning. However, such similarities could also be the result of descent from a common ancestor. In the framework of iterated learning, language evolution converges to an equilibrium that is independent of its starting point, with the effect of shared ancestry decaying over time. Therefore, the central q...

متن کامل

Cultural Transmission and Inductive Biases in Populations of Bayesian Learners

Recent research on computational models of language change and cultural evolution in general has focused on the analytical study of languages as dynamic systems, thus avoiding the difficulties of analysing the complex multi-agent interactions underlying numerical simulations of cultural transmission. The same is true for the examination of the effects of inductive biases on language distributio...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005